Skip to main content

Connect to Destination

This section details the steps to connect to a storage destination where ingested data will be written.

Overview

By default, ingested data is written to the customer's cloud storage account in Azure, AWS S3 or Google Cloud. DataStori follows Medallion Architecture and places ingested data in the following folders:

  1. Bronze folder: Incoming 'as is' raw data is written here.
  2. Silver folder: Deduplicated, flattened, and quality checked data is written into the Silver folder in the Delta format.
  3. Gold folder: Business level aggregated data resides in the Gold folder. DataStori does not directly write into the Gold folder.

Medallion Architecture

Steps

  1. In the Destinations tab, select '+ Add New Destination'.

  2. Select your data storage destination and fill out the connection form with your authentication and credentials. Destination credentials are encrypted and stored in the DataStori application.

Example: Azure SQL Connection Form Sources Page

  1. If your destination is not in the list, please write to contact@datastori.io and we will set it up from the back-end.
    danger

    DataStori cannot validate the destination credentials entered by a user.

info

If your database runs in a private subnet, please ensure that it is accessible to the servers running the data pipelines. The database does not need to be public or open to external tools.

Users can revoke or delete their connections at any time.

File Formats and Database Destinations

File Formats

In addition to the Delta format, DataStori can generate files in other formats including:

  1. Parquet
  2. Iceberg
  3. CSV

Database Destinations

DataStori can write to all the SQL Alchemy supported destinations listed here. Common destinations include:

  1. MySQL
  2. PostgreSQL
  3. Azure SQL
  4. Snowflake
  5. Microsoft SQL Server